Goto

Collaborating Authors

 concept art


Illustrating Classic Brazilian Books using a Text-To-Image Diffusion Model

Mahlow, Felipe, Zanella, André Felipe, Castañeda, William Alberto Cruz, Sarzi-Ribeiro, Regilene Aparecida

arXiv.org Artificial Intelligence

In recent years, Generative Artificial Intelligence (GenAI) has undergone a profound transformation in addressing intricate tasks involving diverse modalities such as textual, auditory, visual, and pictorial generation. Within this spectrum, text-to-image (TTI) models have emerged as a formidable approach to generating varied and aesthetically appealing compositions, spanning applications from artistic creation to realistic facial synthesis, and demonstrating significant advancements in computer vision, image processing, and multimodal tasks. The advent of Latent Diffusion Models (LDMs) signifies a paradigm shift in the domain of AI capabilities. This article delves into the feasibility of employing the Stable Diffusion LDM to illustrate literary works. For this exploration, seven classic Brazilian books have been selected as case studies. The objective is to ascertain the practicality of this endeavor and to evaluate the potential of Stable Diffusion in producing illustrations that augment and enrich the reader's experience. We will outline the beneficial aspects, such as the capacity to generate distinctive and contextually pertinent images, as well as the drawbacks, including any shortcomings in faithfully capturing the essence of intricate literary depictions. Through this study, we aim to provide a comprehensive assessment of the viability and efficacy of utilizing AI-generated illustrations in literary contexts, elucidating both the prospects and challenges encountered in this pioneering application of technology.


Dynamic Prompt Optimizing for Text-to-Image Generation

Mo, Wenyi, Zhang, Tianyu, Bai, Yalong, Su, Bing, Wen, Ji-Rong, Yang, Qing

arXiv.org Artificial Intelligence

Text-to-image generative models, specifically those based on diffusion models like Imagen and Stable Diffusion, have made substantial advancements. Recently, there has been a surge of interest in the delicate refinement of text prompts. Users assign weights or alter the injection time steps of certain words in the text prompts to improve the quality of generated images. However, the success of fine-control prompts depends on the accuracy of the text prompts and the careful selection of weights and time steps, which requires significant manual intervention. To address this, we introduce the \textbf{P}rompt \textbf{A}uto-\textbf{E}diting (PAE) method. Besides refining the original prompts for image generation, we further employ an online reinforcement learning strategy to explore the weights and injection time steps of each word, leading to the dynamic fine-control prompts. The reward function during training encourages the model to consider aesthetic score, semantic consistency, and user preferences. Experimental results demonstrate that our proposed method effectively improves the original prompts, generating visually more appealing images while maintaining semantic alignment. Code is available at https://github.com/Mowenyii/PAE.


Collaborative Generative AI: Integrating GPT-k for Efficient Editing in Text-to-Image Generation

Zhu, Wanrong, Wang, Xinyi, Lu, Yujie, Fu, Tsu-Jui, Wang, Xin Eric, Eckstein, Miguel, Wang, William Yang

arXiv.org Artificial Intelligence

The field of text-to-image (T2I) generation has garnered significant attention both within the research community and among everyday users. Despite the advancements of T2I models, a common issue encountered by users is the need for repetitive editing of input prompts in order to receive a satisfactory image, which is time-consuming and labor-intensive. Given the demonstrated text generation power of large-scale language models, such as GPT-k, we investigate the potential of utilizing such models to improve the prompt editing process for T2I generation. We conduct a series of experiments to compare the common edits made by humans and GPT-k, evaluate the performance of GPT-k in prompting T2I, and examine factors that may influence this process. We found that GPT-k models focus more on inserting modifiers while humans tend to replace words and phrases, which includes changes to the subject matter. Experimental results show that GPT-k are more effective in adjusting modifiers rather than predicting spontaneous changes in the primary subject matters. Adopting the edit suggested by GPT-k models may reduce the percentage of remaining edits by 20-30%.


Alpha Metaverse Technologies Announces AI Tool for Metaverse Content Production

#artificialintelligence

Alpha Metaverse Technologies a technology company that focuses on emerging markets in gaming, NFTs, blockchain gaming, esports, ecommerce, metaverse development and other high growth opportunities. Alpha Metaverse Technologies Inc. is proud to announce the release of our latest AI tool that enables companies to generate concept art to create 3D assets for metaverse experiences at a significantly reduced cost and time. As companies increasingly experiment with metaverse content production, the cost of creating 3D assets has become a major challenge. To address this challenge, Alpha Metaverse Technologies has developed an AI pipeline and automation tool that will enable companies to create concept art for 3D assets quickly and cost-effectively. "Our new AI tool is a game-changer for companies seeking to produce metaverse content," said Brian Wilneff, CEO of Alpha Metaverse.


Building Prey: how artists shape video game worlds

The Guardian

Think of the most impressive, memorable video games of the last 40 years and they tend to have one thing in common: unity of vision. From the sludgy corridors of Doom to the vast art deco chambers of Bioshock, great games take place in intricately realised worlds where every aspect – from armour to architecture – reflects a consistent visual theme. As in the movie industry, the creation of detailed virtual worlds often involves the production of concept art – reams of sketches and paintings, based on early script drafts and discussions, fashioned to provide a target look for designers, artists and coders. "The idea is to establish the big picture of the game, like a visual pitch, not only for artists and designers, but pretty much everyone on the team," says Emmanuel Petit, lead visual designer at Arkane Studios. "I believe this helps in keeping us energised, and gives us something concrete and exciting to work towards."


GDC will host its own video game art gallery next year

Engadget

E3 has the Into the Pixel art gallery for video game concept art, and now the annual Game Developers Conference is trying its hand at one. Called "Art Boss," it's a collaboration with the folks at iam8bit and will feature not just concept art (a la Into the Pixel), but production art, in-game stuff like textures and even promo materials. Another way this differs from Into the Pixel is that each artist will have a headshot and a bio alongside their submitted work -- the E3 gallery typically just lists the artist's name, game a piece came from and the studio responsible. More than that, iam8bit says it's creating a permanent digital and physical archive for the featured works. As for the latter, you'll very likely be able to catch a glimpse at the outfit's locations in Los Angeles and London.


Concept art for the 'Power Ranger' movie's Alpha 5 reveals an alien-looking new design

Los Angeles Times

It's no secret that "Power Rangers" is a departure from the original TV series, and the latest concept art for the movie's new Alpha 5 is further proof. Even fans who thought the updated looks for Rita, the Rangers' suits and the Zords were adequate mental preparation for any other redesigns were likely surprised by Alpha's new design. To borrow some words from Alpha: "Ay yi yi yi yi." Revealed by IGN, this new take on the Power Ranger ally is more than a bit of a departure from the character's original look. In the TV series, Alpha was Zordon's panicky but devoted robotic assistant who served as the Rangers' adviser and friend. While the new Alpha retains the gold saucer-shaped helmet as well as the red body, not much else is recognizable.